In this work, we perform a wide variety of experiments with different deep learning architectures on datasets of limited size. According to our study, we show that model complexity is a critical factor when only a few samples per class are available. Differently from the literature, we show that in some configurations, the state of the art can be improved using low complexity models. For instance, in problems with scarce training samples and without data augmentation, low-complexity convolutional neural networks perform comparably well or better than state-of-the-art architectures. Moreover, we show that even standard data augmentation can boost recognition performance by large margins. This result suggests the development of more complex data generation/augmentation pipelines for cases when data is limited. Finally, we show that dropout, a widely used regularization technique, maintains its role as a good regularizer even when data is scarce. Our findings are empirically validated on the sub-sampled versions of popular CIFAR-10, Fashion-MNIST and, SVHN benchmarks.

A Close Look at Deep Learning with Small Data / Brigato, Lorenzo; Iocchi, Luca. - (2020), pp. 2490-2497. (Intervento presentato al convegno International Conference on Pattern Recognition tenutosi a Milano) [10.1109/icpr48806.2021.9412492].

A Close Look at Deep Learning with Small Data

Brigato, Lorenzo
;
Iocchi, Luca
2020

Abstract

In this work, we perform a wide variety of experiments with different deep learning architectures on datasets of limited size. According to our study, we show that model complexity is a critical factor when only a few samples per class are available. Differently from the literature, we show that in some configurations, the state of the art can be improved using low complexity models. For instance, in problems with scarce training samples and without data augmentation, low-complexity convolutional neural networks perform comparably well or better than state-of-the-art architectures. Moreover, we show that even standard data augmentation can boost recognition performance by large margins. This result suggests the development of more complex data generation/augmentation pipelines for cases when data is limited. Finally, we show that dropout, a widely used regularization technique, maintains its role as a good regularizer even when data is scarce. Our findings are empirically validated on the sub-sampled versions of popular CIFAR-10, Fashion-MNIST and, SVHN benchmarks.
2020
International Conference on Pattern Recognition
image classification; small data-set; banchmarking
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
A Close Look at Deep Learning with Small Data / Brigato, Lorenzo; Iocchi, Luca. - (2020), pp. 2490-2497. (Intervento presentato al convegno International Conference on Pattern Recognition tenutosi a Milano) [10.1109/icpr48806.2021.9412492].
File allegati a questo prodotto
File Dimensione Formato  
Brigato_A-Close_2020.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 726.86 kB
Formato Adobe PDF
726.86 kB Adobe PDF   Contatta l'autore
Brigato_preprint_A-Close_2020.pdf

accesso aperto

Note: DOI10.1109/ICPR48806.2021.9412492
Tipologia: Documento in Pre-print (manoscritto inviato all'editore, precedente alla peer review)
Licenza: Creative commons
Dimensione 669.9 kB
Formato Adobe PDF
669.9 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1705180
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 82
  • ???jsp.display-item.citation.isi??? 66
social impact